An Adaptive Test of Independence with Analytic Kernel Embeddings
ثبت نشده
چکیده
B. Redundant Test Locations Here, we provide a simple illustration to show that two locations t1 = (v1,w1) and t2 = (v2,w2) which are too close to each other will reduce the optimization objective. We consider the Sinusoid problem described in Section 3.1 with ω = 1, and use J = 2 test locations. In Figure 6, t1 is fixed at the red star, while t2 is varied along the horizontal line. The objective value λ̂n as a function of t2 is shown in the bottom figure. It can be seen that λ̂n decreases sharply when t2 is in the neighborhood of t1. This property implies that two locations which are too close will not maximize the objective function (i.e., the second feature contains no additional information when it matches the first). For J > 2, the objective sharply decreases if any two locations are in the same neighborhood.
منابع مشابه
An Adaptive Test of Independence with Analytic Kernel Embeddings
A new computationally efficient dependence measure, and an adaptive statistical test of independence, are proposed. The dependence measure is the difference between analytic embeddings of the joint distribution and the product of the marginals, evaluated at a finite set of locations (features). These features are chosen so as to maximize a lower bound on the test power, resulting in a test that...
متن کاملA Wild Bootstrap for Degenerate Kernel Tests
A wild bootstrap method for nonparametric hypothesis tests based on kernel distribution embeddings is proposed. This bootstrap method is used to construct provably consistent tests that apply to random processes, for which the naive permutation-based bootstrap fails. It applies to a large group of kernel tests based on V-statistics, which are degenerate under the null hypothesis, and nondegener...
متن کاملA Kernel Test for Three-Variable Interactions
We introduce kernel nonparametric tests for Lancaster three-variable interaction and for total independence, using embeddings of signed measures into a reproducing kernel Hilbert space. The resulting test statistics are straightforward to compute, and are used in powerful interaction tests, which are consistent against all alternatives for a large family of reproducing kernels. We show the Lanc...
متن کاملKernel Measures of Independence for non-iid Data
Many machine learning algorithms can be formulated in the framework of statistical independence such as the Hilbert Schmidt Independence Criterion. In this paper, we extend this criterion to deal with structured and interdependent observations. This is achieved by modeling the structures using undirected graphical models and comparing the Hilbert space embeddings of distributions. We apply this...
متن کاملCharacteristic and Universal Tensor Product Kernels
Kernel mean embeddings provide a versatile and powerful nonparametric representation of probability distributions with several fundamental applications in machine learning. Key to the success of the technique is whether the embedding is injective. This characteristic property of the underlying kernel ensures that probability distributions can be discriminated via their representations. In this ...
متن کامل